# 1.58-bit quantization

Bitnet B1.58 2B 4T GGUF
MIT
A 1.58-bit quantized large language model developed by Microsoft, designed for efficient inference, offering IQ2_BN and IQ2_BN_R4 quantization versions
Large Language Model
B
tdh111
1,058
4
Falcon E 3B Instruct
Other
Falcon-E-3B-Instruct is an efficient language model based on a 1.58-bit architecture, optimized for edge devices, with excellent inference capabilities and low memory usage.
Large Language Model Transformers
F
tiiuae
225
22
Falcon E 1B Instruct
Other
Falcon-E-1B-Instruct is an efficient language model based on a 1.58-bit architecture, optimized for edge devices with low memory footprint and high performance.
Large Language Model Transformers
F
tiiuae
87
7
Falcon E 3B Base
Other
Falcon-E is a 1.58-bit quantized language model developed by TII, featuring a pure Transformer architecture designed for efficient inference
Large Language Model Transformers
F
tiiuae
51
6
Bitnet B1.58 2B 4T Gguf
MIT
The first open-source, native 1-bit large language model developed by Microsoft Research, with a parameter scale of 2 billion, trained on a corpus of 4 trillion tokens.
Large Language Model English
B
microsoft
25.77k
143
Bitnet B1.58 2B 4T
MIT
The first open-source 2-billion-parameter native 1-bit large language model developed by Microsoft Research, trained on 4 trillion tokens, demonstrating that native 1-bit large language models can significantly improve computational efficiency while maintaining performance comparable to full-precision open-source models of the same scale.
Large Language Model Transformers English
B
microsoft
35.87k
846
Bitnet B1.58 2B 4T Bf16
MIT
An open-source native 1-bit large language model developed by Microsoft Research, with 2 billion parameters trained on a 4 trillion token corpus, significantly improving computational efficiency.
Large Language Model Transformers English
B
microsoft
2,968
24
Falcon E 1B Base
Other
Falcon-E-1B-Base is an efficient 1.58-bit language model developed by TII, featuring a pure Transformer architecture and optimized for edge devices.
Large Language Model Transformers
F
tiiuae
53
4
Llama3 8B 1.58 100B Tokens GGUF
A GGUF format model converted from Meta-Llama-3-8B-Instruct and HF1BitLLM/Llama3-8B-1.58-100B-tokens models, suitable for llama.cpp inference
Large Language Model Transformers
L
brunopio
2,035
16
Llama3 8B 1.58 100B Tokens
Large language model fine-tuned based on BitNet 1.58b architecture, with Llama-3-8B-Instruct as the base model, employing extreme quantization techniques
Large Language Model Transformers
L
HF1BitLLM
2,427
181
Bitnet B1 58 Xl Q8 0 Gguf
MIT
BitNet b1.58 is a large language model with 1.58-bit quantization. It reduces the computational resource requirements by lowering the weight precision while maintaining performance close to that of a full-precision model.
Large Language Model Transformers
B
BoscoTheDog
326
7
Bitnet B1 58 Large
MIT
BitNet b1.58 is a 1-bit large language model with 3 billion parameters, trained on the RedPajama dataset for 100 billion tokens.
Large Language Model Transformers
B
1bitLLM
10.17k
95
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase